Fars Province
ExBigBang: A Dynamic Approach for Explainable Persona Classification through Contextualized Hybrid Transformer Analysis
Afzoon, Saleh, Beheshti, Amin, Rezvani, Nabi, Khunjush, Farshad, Naseem, Usman, McMahon, John, Fathollahi, Zahra, Labani, Mahdieh, Mansoor, Wathiq, Zhang, Xuyun
In user-centric design, persona development plays a vital role in understanding user behaviour, capturing needs, segmenting audiences, and guiding design decisions. However, the growing complexity of user interactions calls for a more contextualized approach to ensure designs align with real user needs. While earlier studies have advanced persona classification by modelling user behaviour, capturing contextual information, especially by integrating textual and tabular data, remains a key challenge. These models also often lack explainability, leaving their predictions difficult to interpret or justify. To address these limitations, we present ExBigBang (Explainable BigBang), a hybrid text-tabular approach that uses transformer-based architectures to model rich contextual features for persona classification. ExBigBang incorporates metadata, domain knowledge, and user profiling to embed deeper context into predictions. Through a cyclical process of user profiling and classification, our approach dynamically updates to reflect evolving user behaviours. Experiments on a benchmark persona classification dataset demonstrate the robustness of our model. An ablation study confirms the benefits of combining text and tabular data, while Explainable AI techniques shed light on the rationale behind the model's predictions.
- Oceania > Australia > New South Wales > Sydney (0.04)
- Asia > Middle East > UAE > Dubai Emirate > Dubai (0.04)
- South America > Argentina > Patagonia > Río Negro Province > Viedma (0.04)
- (10 more...)
- Information Technology > Security & Privacy (0.93)
- Education (0.93)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.46)
Efficient Triple Modular Redundancy for Reliability Enhancement of DNNs Using Explainable AI
Soroush, Kimia, Shirazi, Nastaran, Raji, Mohsen
Deep Neural Networks (DNNs) are widely employed in safety-critical domains, where ensuring their reliability is essential. Triple Modular Redundancy (TMR) is an effective technique to enhance the reliability of DNNs in the presence of bit-flip faults. In order to handle the significant overhead of TMR, it is applied selectively on the parameters and components with the highest contribution at the model output. Hence, the accuracy of the selection criterion plays the key role on the efficiency of TMR. This paper presents an efficient TMR approach to enhance the reliability of DNNs against bit-flip faults using an Explainable Artificial Intelligence (XAI) method. Since XAI can provide valuable insights about the importance of individual neurons and weights in the performance of the network, they can be applied as the selection metric in TMR techniques. The proposed method utilizes a low-cost, gradient-based XAI technique known as Layer-wise Relevance Propagation (LRP) to calculate importance scores for DNN parameters. These scores are then used to enhance the reliability of the model, with the most critical weights being protected by TMR. The proposed approach is evaluated on two DNN models, VGG16 and AlexNet, using datasets such as MNIST and CIFAR-10. The results demonstrate that the method can protect the AlexNet model at a bit error rate of 10-4, achieving over 60% reliability improvement while maintaining the same overhead as state-of-the-art methods.
Compressing Deep Neural Networks Using Explainable AI
Soroush, Kimia, Raji, Mohsen, Ghavami, Behnam
Deep neural networks (DNNs) have demonstrated remarkable performance in many tasks but it often comes at a high computational cost and memory usage. Compression techniques, such as pruning and quantization, are applied to reduce the memory footprint of DNNs and make it possible to accommodate them on resource-constrained edge devices. Recently, explainable artificial intelligence (XAI) methods have been introduced with the purpose of understanding and explaining AI methods. XAI can be utilized to get to know the inner functioning of DNNs, such as the importance of different neurons and features in the overall performance of DNNs. In this paper, a novel DNN compression approach using XAI is proposed to efficiently reduce the DNN model size with negligible accuracy loss. In the proposed approach, the importance score of DNN parameters (i.e. weights) are computed using a gradient-based XAI technique called Layer-wise Relevance Propagation (LRP). Then, the scores are used to compress the DNN as follows: 1) the parameters with the negative or zero importance scores are pruned and removed from the model, 2) mixed-precision quantization is applied to quantize the weights with higher/lower score with higher/lower number of bits. The experimental results show that, the proposed compression approach reduces the model size by 64% while the accuracy is improved by 42% compared to the state-of-the-art XAI-based compression method.
- Asia > Middle East > Iran > Fars Province > Shiraz (0.04)
- Asia > Middle East > Iran > Kerman Province > Kerman (0.04)
- Africa > Mozambique > Gaza Province > Xai-Xai (0.04)
Containment Control Approach for Steering Opinion in a Social Network
The paper studies the problem of steering multi-dimensional opinion in a social network. Assuming the society of desire consists of stubborn and regular agents, stubborn agents are considered as leaders who specify the desired opinion distribution as a distributed reward or utility function. In this context, each regular agent is seen as a follower, updating its bias on the initial opinion and influence weights by averaging their observations of the rewards their influencers have received. Assuming random graphs with reducible and irreducible topology specify the influences on regular agents, opinion evolution is represented as a containment control problem in which stability and convergence to the final opinion are proven.
- North America > United States > Florida > Orange County > Orlando (0.14)
- North America > United States > Arizona > Pima County > Tucson (0.14)
- Asia > Middle East > Jordan (0.04)
- (2 more...)
Texture Image Synthesis Using Spatial GAN Based on Vision Transformers
Salari, Elahe, Azimifar, Zohreh
Texture synthesis is a fundamental task in computer vision, whose goal is to generate visually realistic and structurally coherent textures for a wide range of applications, from graphics to scientific simulations. While traditional methods like tiling and patch-based techniques often struggle with complex textures, recent advancements in deep learning have transformed this field. In this paper, we propose ViT-SGAN, a new hybrid model that fuses Vision Transformers (ViTs) with a Spatial Generative Adversarial Network (SGAN) to address the limitations of previous methods. By incorporating specialized texture descriptors such as mean-variance (mu, sigma) and textons into the self-attention mechanism of ViTs, our model achieves superior texture synthesis. This approach enhances the model's capacity to capture complex spatial dependencies, leading to improved texture quality that is superior to state-of-the-art models, especially for regular and irregular textures. Comparison experiments with metrics such as FID, IS, SSIM, and LPIPS demonstrate the substantial improvement of ViT-SGAN, which underlines its efficiency in generating diverse realistic textures.
Enhancing Osteoporosis Detection: An Explainable Multi-Modal Learning Framework with Feature Fusion and Variable Clustering
Chagahi, Mehdi Hosseini, Dashtaki, Saeed Mohammadi, Delfan, Niloufar, Mohammadi, Nadia, Samari, Alireza, Moshiri, Behzad, Piran, Md. Jalil, Faust, Oliver
Osteoporosis is a common condition that increases fracture risk, especially in older adults. Early diagnosis is vital for preventing fractures, reducing treatment costs, and preserving mobility. However, healthcare providers face challenges like limited labeled data and difficulties in processing medical images. This study presents a novel multi-modal learning framework that integrates clinical and imaging data to improve diagnostic accuracy and model interpretability. The model utilizes three pre-trained networks-VGG19, InceptionV3, and ResNet50-to extract deep features from X-ray images. These features are transformed using PCA to reduce dimensionality and focus on the most relevant components. A clustering-based selection process identifies the most representative components, which are then combined with preprocessed clinical data and processed through a fully connected network (FCN) for final classification. A feature importance plot highlights key variables, showing that Medical History, BMI, and Height were the main contributors, emphasizing the significance of patient-specific data. While imaging features were valuable, they had lower importance, indicating that clinical data are crucial for accurate predictions. This framework promotes precise and interpretable predictions, enhancing transparency and building trust in AI-driven diagnoses for clinical integration.
- Asia > Middle East > Iran > Tehran Province > Tehran (0.05)
- North America > United States > New Mexico > Bernalillo County > Albuquerque (0.04)
- North America > Canada (0.04)
- (4 more...)
- Health & Medicine > Therapeutic Area > Musculoskeletal (1.00)
- Health & Medicine > Therapeutic Area > Rheumatology (0.95)
- Health & Medicine > Diagnostic Medicine > Imaging (0.89)
PersoBench: Benchmarking Personalized Response Generation in Large Language Models
Afzoon, Saleh, Naseem, Usman, Beheshti, Amin, Jamali, Zahra
Large Language Models (LLMs) have revolutionized NLP, excelling in human-like text generation across domains and becoming central to dialogue systems. However, evaluating their ability to generate personalized responses that enhance user engagement is crucial, especially in applications like customer service, where tailored interactions boost satisfaction [1]. While recent benchmarks such as RPBench-Auto [2], TIMECHARA [3] and RoleLLM [4] have been introduced in the role-playing domain to assess LLMs' adherence to predefined characters or roles in character-based, scene-based, and temporal setups, there is still no dedicated benchmark for automatic personalized response generation of LLMs in the literature. Further, existing benchmarks also suffer from biases in their evaluations due to the use of large LLMs as judges, and limited experimental sizes constrain them. To fill this gap, we introduce PersoBench, a benchmark for response personalization, to assess the strengths and limitations of current LLMs in generating personalized responses. To the best of our knowledge, no prior work has introduced a comprehensive benchmark specifically focused on evaluating response personalization in LLMs. Using comprehensive datasets and a diverse set of established metrics, including fluency, diversity, and coherence, we ensure a robust evaluation of various aspects of response generation, drawing on insights from a recent survey in the field [1]. More specifically, in line with this objective of the mentioned context, we aim to answer the following research questions: 1. Can LLMs generate fluent responses?
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > United States > Michigan > Washtenaw County > Ann Arbor (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- (5 more...)
TransDAE: Dual Attention Mechanism in a Hierarchical Transformer for Efficient Medical Image Segmentation
Azad, Bobby, Adibfar, Pourya, Fu, Kaiqun
In healthcare, medical image segmentation is crucial for accurate disease diagnosis and the development of effective treatment strategies. Early detection can significantly aid in managing diseases and potentially prevent their progression. Machine learning, particularly deep convolutional neural networks, has emerged as a promising approach to addressing segmentation challenges. Traditional methods like U-Net use encoding blocks for local representation modeling and decoding blocks to uncover semantic relationships. However, these models often struggle with multi-scale objects exhibiting significant variations in texture and shape, and they frequently fail to capture long-range dependencies in the input data. Transformers designed for sequence-to-sequence predictions have been proposed as alternatives, utilizing global self-attention mechanisms. Yet, they can sometimes lack precise localization due to insufficient granular details. To overcome these limitations, we introduce TransDAE: a novel approach that reimagines the self-attention mechanism to include both spatial and channel-wise associations across the entire feature space, while maintaining computational efficiency. Additionally, TransDAE enhances the skip connection pathway with an inter-scale interaction module, promoting feature reuse and improving localization accuracy. Remarkably, TransDAE outperforms existing state-of-the-art methods on the Synaps multi-organ dataset, even without relying on pre-trained weights.
- North America > United States > South Dakota (0.05)
- Europe > Switzerland (0.04)
- Europe > France > Grand Est > Bas-Rhin > Strasbourg (0.04)
- Asia > Middle East > Iran > Fars Province > Shiraz (0.04)
- Health & Medicine > Therapeutic Area (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
Large Language Models versus Classical Machine Learning: Performance in COVID-19 Mortality Prediction Using High-Dimensional Tabular Data
Ghaffarzadeh-Esfahani, Mohammadreza, Ghaffarzadeh-Esfahani, Mahdi, Salahi-Niri, Arian, Toreyhi, Hossein, Atf, Zahra, Mohsenzadeh-Kermani, Amirali, Sarikhani, Mahshad, Tajabadi, Zohreh, Shojaeian, Fatemeh, Bagheri, Mohammad Hassan, Feyzi, Aydin, Tarighatpayma, Mohammadamin, Gazmeh, Narges, Heydari, Fateme, Afshar, Hossein, Allahgholipour, Amirreza, Alimardani, Farid, Salehi, Ameneh, Asadimanesh, Naghmeh, Khalafi, Mohammad Amin, Shabanipour, Hadis, Moradi, Ali, Zadeh, Sajjad Hossein, Yazdani, Omid, Esbati, Romina, Maleki, Moozhan, Nasr, Danial Samiei, Soheili, Amirali, Majlesi, Hossein, Shahsavan, Saba, Soheilipour, Alireza, Goudarzi, Nooshin, Taherifard, Erfan, Hatamabadi, Hamidreza, Samaan, Jamil S, Savage, Thomas, Sakhuja, Ankit, Soroush, Ali, Nadkarni, Girish, Darazam, Ilad Alavi, Pourhoseingholi, Mohamad Amin, Safavi-Naini, Seyed Amir Ahmad
Background: This study aimed to evaluate and compare the performance of classical machine learning models (CMLs) and large language models (LLMs) in predicting mortality associated with COVID-19 by utilizing a high-dimensional tabular dataset. Materials and Methods: We analyzed data from 9,134 COVID-19 patients collected across four hospitals. Seven CML models, including XGBoost and random forest (RF), were trained and evaluated. The structured data was converted into text for zero-shot classification by eight LLMs, including GPT-4 and Mistral-7b. Additionally, Mistral-7b was fine-tuned using the QLoRA approach to enhance its predictive capabilities. Results: Among the CML models, XGBoost and RF achieved the highest accuracy, with F1 scores of 0.87 for internal validation and 0.83 for external validation. In the LLM category, GPT-4 was the top performer with an F1 score of 0.43. Fine-tuning Mistral-7b significantly improved its recall from 1% to 79%, resulting in an F1 score of 0.74, which was stable during external validation. Conclusion: While LLMs show moderate performance in zero-shot classification, fine-tuning can significantly enhance their effectiveness, potentially aligning them closer to CML models. However, CMLs still outperform LLMs in high-dimensional tabular data tasks.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- Europe > United Kingdom > England > Nottinghamshire > Nottingham (0.14)
- Asia > Middle East > Iran > Tehran Province > Tehran (0.08)
- (8 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Quantizing YOLOv7: A Comprehensive Study
Baghbanbashi, Mohammadamin, Raji, Mohsen, Ghavami, Behnam
YOLO is a deep neural network (DNN) model presented for robust real-time object detection following the one-stage inference approach. It outperforms other real-time object detectors in terms of speed and accuracy by a wide margin. Nevertheless, since YOLO is developed upon a DNN backbone with numerous parameters, it will cause excessive memory load, thereby deploying it on memory-constrained devices is a severe challenge in practice. To overcome this limitation, model compression techniques, such as quantizing parameters to lower-precision values, can be adopted. As the most recent version of YOLO, YOLOv7 achieves such state-of-the-art performance in speed and accuracy in the range of 5 FPS to 160 FPS that it surpasses all former versions of YOLO and other existing models in this regard. So far, the robustness of several quantization schemes has been evaluated on older versions of YOLO. These methods may not necessarily yield similar results for YOLOv7 as it utilizes a different architecture. In this paper, we conduct in-depth research on the effectiveness of a variety of quantization schemes on the pre-trained weights of the state-of-the-art YOLOv7 model. Experimental results demonstrate that using 4-bit quantization coupled with the combination of different granularities results in ~3.92x and ~3.86x memory-saving for uniform and non-uniform quantization, respectively, with only 2.5% and 1% accuracy loss compared to the full-precision baseline model.
- Asia > Middle East > Iran > Fars Province > Shiraz (0.05)
- North America > United States > Florida > Palm Beach County > Boca Raton (0.04)
- Asia > Middle East > Iran > Kerman Province > Kerman (0.04)